# Multi-task Fine-tuning

Tooka SBERT V2 Large
A semantic text similarity and embedding model specifically designed for Persian, capable of mapping sentences into a dense vector space where semantically similar texts are positioned close to each other.
Text Embedding
T
PartAI
127
1
Olmo 2 0425 1B SFT
Apache-2.0
OLMo 2 1B SFT is a supervised fine-tuned version of the OLMo-2-0425-1B model, trained on the Tulu 3 dataset, designed to achieve state-of-the-art performance across multiple tasks.
Large Language Model Transformers English
O
allenai
1,759
2
DPO A5 Nlp
TRL is a reinforcement learning library based on the Transformer architecture for training and fine-tuning language models.
Large Language Model Transformers
D
EraCoding
26
1
Paligemma2 3b Mix 448
PaliGemma 2 is a vision-language model based on Gemma 2, supporting image and text inputs with text generation output, suitable for various vision-language tasks.
Image-to-Text Transformers
P
google
20.55k
44
Camembertav2 Base
MIT
CamemBERTav2 is a French language model pretrained on 275 billion French text tokens, utilizing the DebertaV2 architecture, and excels in multiple French NLP tasks.
Large Language Model Transformers French
C
almanach
2,972
19
Phico D Instruck
MIT
An instruction-following model fine-tuned based on T5-base, specifically designed to understand and execute complex instructions
Large Language Model Transformers Supports Multiple Languages
P
acecalisto3
19
2
Paligemma 3b Ft Refcoco Seg 896
PaliGemma is a lightweight vision-language model developed by Google, built upon the SigLIP vision model and Gemma language model, supporting multilingual text generation and visual understanding tasks.
Image-to-Text Transformers
P
google
20
6
Tookabert Base
Apache-2.0
TookaBERT is a family of encoder models trained on Persian, including base and large versions, suitable for various natural language processing tasks.
Large Language Model Transformers Other
T
PartAI
127
24
Openelm 3B
OpenELM is a set of open-source efficient language models that employ a hierarchical scaling strategy to optimize parameter allocation and improve model accuracy. It includes four parameter scales: 270M, 450M, 1.1B, and 3B, offering both pre-trained and instruction-tuned versions.
Large Language Model Transformers
O
apple
1,436
123
Hyperion 2.0 Mistral 7B
Apache-2.0
A multi-domain language model fine-tuned on the Hyperion-v2.0 dataset, excelling in scientific reasoning and complex task processing.
Large Language Model Transformers Supports Multiple Languages
H
Locutusque
16
6
Kafkalm 70B German V0.1
A large German language model developed based on Llama2 70B, specializing in German business scenarios
Large Language Model Transformers German
K
seedboxai
159
18
H2o Danube 1.8b Base
Apache-2.0
A 1.8B parameter base language model trained by H2O.ai, based on an improved Llama 2 architecture with 16K context length support
Large Language Model Transformers English
H
h2oai
281
43
Mental Alpaca
This is a fine-tuned large language model for mental health prediction using online text data.
Large Language Model Transformers English
M
NEU-HAI
180
9
Ziya LLaMA 13B V1
Gpl-3.0
A 13-billion-parameter pre-trained model based on the LLaMa architecture, capable of translation, programming, text classification, information extraction, summarization, copywriting, commonsense Q&A, and mathematical calculations
Large Language Model Transformers Supports Multiple Languages
Z
IDEA-CCNL
219
275
All Mpnet Base V2
Apache-2.0
Sentence embedding model based on MPNet architecture, mapping text to a 768-dimensional vector space, suitable for semantic search and text similarity tasks
Text Embedding English
A
diptanuc
138
1
Glm 2b
GLM-2B is a general-purpose language model pre-trained with autoregressive blank filling objectives, supporting various natural language understanding and generation tasks.
Large Language Model Transformers English
G
THUDM
60
16
Deberta Base Combined Squad1 Aqa Newsqa And Newsqa
MIT
This model is a question-answering model based on the DeBERTa architecture, fine-tuned on the SQuAD1, AQA, and NewsQA datasets.
Question Answering System Transformers
D
stevemobs
15
0
Erlangshen MegatronBert 1.3B Sentiment
Apache-2.0
A Chinese sentiment analysis model based on the MegatronBert architecture, fine-tuned on multiple sentiment analysis tasks
Text Classification Transformers Chinese
E
IDEA-CCNL
93
20
Erlangshen Roberta 330M NLI
Apache-2.0
A fine-tuned version based on the Chinese RoBERTa-wwm-ext-large model on multiple natural language inference task datasets
Text Classification Transformers Chinese
E
IDEA-CCNL
468
5
Distilbert Base Uncased Squad2 With Ner With Neg With Multi With Repeat
A QA and NER model fine-tuned on the conll2003 dataset based on distilbert-base-uncased-squad2
Question Answering System Transformers
D
andi611
20
0
Tapas Large Finetuned Wtq
Apache-2.0
TAPAS is a table question answering model based on the BERT architecture, pre-trained in a self-supervised manner on Wikipedia table data, supporting natural language question answering on table content
Question Answering System Transformers English
T
google
124.85k
141
Robbert V2 Dutch Ner
MIT
RobBERT is the state-of-the-art Dutch BERT model, pretrained on a large scale and adaptable to various text tasks through fine-tuning.
Large Language Model Other
R
pdelobelle
76.94k
3
Data2vec Nlp Base
Apache-2.0
Data2Vec NLP Base is a natural language processing model converted from the fairseq framework, suitable for tasks such as text classification.
Large Language Model Transformers
D
edugp
14
0
Sew D Mid 400k Ft Ls100h
Apache-2.0
SEW-D-mid is a speech pre-training model developed by ASAPP Research, focusing on automatic speech recognition tasks, achieving a good balance between performance and efficiency.
Speech Recognition Transformers English
S
asapp
20
1
Tapas Medium Finetuned Wikisql Supervised
Apache-2.0
TAPAS is a Transformer-based table question answering model, pre-trained in a self-supervised manner on English Wikipedia table data and fine-tuned with supervision on the WikiSQL dataset.
Question Answering System Transformers English
T
google
19
0
T5 V1 1 Large
Apache-2.0
T5 1.1 is Google's improved text-to-text transfer model, utilizing GEGLU activation function and optimized architecture, focusing on unsupervised pre-training
Large Language Model English
T
google
111.29k
17
Bertino
MIT
A lightweight DistilBERT model pre-trained on a large-scale Italian corpus, suitable for various natural language processing tasks.
Large Language Model Transformers Other
B
indigo-ai
64
16
Tapas Base Finetuned Wikisql Supervised
Apache-2.0
TAPAS is a BERT-based Transformer model specifically designed for table question answering tasks. It is pre-trained in a self-supervised manner on English Wikipedia table data and supports weakly supervised table parsing.
Question Answering System Transformers English
T
google
737
9
S Biomed Roberta Snli Multinli Stsb
This is a sentence transformer model based on allenai/biomed_roberta_base, specifically fine-tuned for sentence similarity tasks, capable of mapping text to a 768-dimensional vector space.
Text Embedding Transformers
S
pritamdeka
270
4
Bart Base
Apache-2.0
BART is a Transformer model combining a bidirectional encoder and an autoregressive decoder, suitable for text generation and understanding tasks.
Large Language Model English
B
facebook
2.1M
183
Minilm L12 H384 Uncased
MIT
MiniLM is a compact and efficient pre-trained language model, compressed through deep self-attention distillation technology, suitable for language understanding and generation tasks.
Large Language Model
M
microsoft
10.19k
89
Electra Large Generator
Apache-2.0
ELECTRA is an efficient self-supervised language representation learning method that replaces traditional generative pretraining with discriminative pretraining, significantly improving computational efficiency.
Large Language Model English
E
google
473
8
Tapas Large Finetuned Wikisql Supervised
Apache-2.0
TAPAS is a BERT-like Transformer model designed for table-based question answering tasks. It is pre-trained in a self-supervised manner on English Wikipedia table corpora and fine-tuned on the WikiSQL dataset.
Question Answering System Transformers English
T
google
80
6
Tapas Small Finetuned Wtq
Apache-2.0
This model is a small version of TAPAS, specifically fine-tuned on the WikiTable Questions dataset for table-based question answering tasks.
Question Answering System Transformers English
T
google
406
5
Tapas Medium Finetuned Wtq
Apache-2.0
This model is a medium-sized table question answering model based on TAPAS architecture, fine-tuned on WikiTable Questions dataset, suitable for table data QA tasks.
Question Answering System Transformers English
T
google
77
2
Albert Fa Base V2
Apache-2.0
A lightweight BERT model for self-supervised learning of Persian language representations
Large Language Model Transformers Other
A
m3hrdadfi
43
4
Tapas Tiny Finetuned Wtq
Apache-2.0
TAPAS is a tiny Transformer model optimized for table question answering tasks, achieving table comprehension capabilities through intermediate pretraining and chained multi-dataset fine-tuning
Question Answering System Transformers English
T
google
1,894
1
Mt5 Small Jaquad Qg Ae
A Japanese question generation and answer extraction model fine-tuned based on MT5-small, supporting generating questions or extracting answers from given text.
Question Answering System Transformers Japanese
M
lmqg
143
1
Wangchanberta Base Wiki Newmm
A RoBERTa BASE model pretrained on Thai Wikipedia, suitable for Thai text processing tasks
Large Language Model Other
W
airesearch
115
2
Tapas Base Finetuned Wtq
Apache-2.0
TAPAS is a Transformer-based table question answering model, pre-trained on Wikipedia table data through self-supervised learning and fine-tuned on datasets like WTQ.
Question Answering System Transformers English
T
google
23.03k
217
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase